Your browser doesn't support javascript.
loading
: 20 | 50 | 100
1 - 20 de 408
1.
Open Forum Infect Dis ; 11(4): ofae113, 2024 Apr.
Article En | MEDLINE | ID: mdl-38560600

Background: Diagnosis of cutaneous leishmaniasis (CL) usually relies on invasive samples, but it is unclear whether more patient-friendly tools are good alternatives for diverse lesions when used with polymerase chain reaction (PCR). Methods: Patients with suspected CL were enrolled consecutively in a prospective diagnostic accuracy study. We compared dental broach, tape disc, and microbiopsy samples with PCR as index tests, using PCR with skin slit samples as reference test. Subsequently, we constructed a composite reference test including microscopy, the 3 index tests and skin slit PCR, and we compared these same tests with the composite reference test. We assessed diagnostic accuracy parameters with 95% confidence intervals for all comparisons. Results: Among 344 included patients, 282 (82.0%) had CL diagnosed, and 62 (18.0%) CL absence, by skin slit PCR. The sensitivity and specificity by PCR were 89.0% (95% confidence interval, 84.8%-92.1%) and 58.1% (45.7%-69.5%), respectively, for dental broach, 96.1% (93.2%-97.8%) and 27.4% (17.9%-39.6%) for tape disc, and 74.8% (66.3%-81.7%) and 72.7% (51.8%-86.8%) for microbiopsy. Several reference test-negative patients were consistently positive with the index tests. Using the composite reference test, dental broach, and skin slit had similar diagnostic performance. Discussion: Dental broach seems a less invasive but similarly accurate alternative to skin slit for diagnosing CL when using PCR. Tape discs lack specificity and seem unsuitable for CL diagnosis without cutoff. Reference tests for CL are problematic, since using a single reference test is likely to miss true cases, while composite reference tests are often biased and impractical as they require multiple tests.

3.
J Clin Epidemiol ; 168: 111270, 2024 Apr.
Article En | MEDLINE | ID: mdl-38311188

OBJECTIVES: To systematically evaluate the performance of COVID-19 prognostic models and scores for mortality risk in older populations across three health-care settings: hospitals, primary care, and nursing homes. STUDY DESIGN AND SETTING: This retrospective external validation study included 14,092 older individuals of ≥70 years of age with a clinical or polymerase chain reaction-confirmed COVID-19 diagnosis from March 2020 to December 2020. The six validation cohorts include three hospital-based (CliniCo, COVID-OLD, COVID-PREDICT), two primary care-based (Julius General Practitioners Network/Academisch network huisartsgeneeskunde/Network of Academic general Practitioners, PHARMO), and one nursing home cohort (YSIS) in the Netherlands. Based on a living systematic review of COVID-19 prediction models using Prediction model Risk Of Bias ASsessment Tool for quality and risk of bias assessment and considering predictor availability in validation cohorts, we selected six prognostic models predicting mortality risk in adults with COVID-19 infection (GAL-COVID-19 mortality, 4C Mortality Score, National Early Warning Score 2-extended model, Xie model, Wang clinical model, and CURB65 score). All six prognostic models were validated in the hospital cohorts and the GAL-COVID-19 mortality model was validated in all three healthcare settings. The primary outcome was in-hospital mortality for hospitals and 28-day mortality for primary care and nursing home settings. Model performance was evaluated in each validation cohort separately in terms of discrimination, calibration, and decision curves. An intercept update was performed in models indicating miscalibration followed by predictive performance re-evaluation. MAIN OUTCOME MEASURE: In-hospital mortality for hospitals and 28-day mortality for primary care and nursing home setting. RESULTS: All six prognostic models performed poorly and showed miscalibration in the older population cohorts. In the hospital settings, model performance ranged from calibration-in-the-large -1.45 to 7.46, calibration slopes 0.24-0.81, and C-statistic 0.55-0.71 with 4C Mortality Score performing as the most discriminative and well-calibrated model. Performance across health-care settings was similar for the GAL-COVID-19 model, with a calibration-in-the-large in the range of -2.35 to -0.15 indicating overestimation, calibration slopes of 0.24-0.81 indicating signs of overfitting, and C-statistic of 0.55-0.71. CONCLUSION: Our results show that most prognostic models for predicting mortality risk performed poorly in the older population with COVID-19, in each health-care setting: hospital, primary care, and nursing home settings. Insights into factors influencing predictive model performance in the older population are needed for pandemic preparedness and reliable prognostication of health-related outcomes in this demographic.


COVID-19 , Adult , Humans , Aged , Prognosis , COVID-19/diagnosis , Retrospective Studies , COVID-19 Testing , Nursing Homes , Hospitals , Hospital Mortality , Primary Health Care
4.
Nat Methods ; 21(2): 182-194, 2024 Feb.
Article En | MEDLINE | ID: mdl-38347140

Validation metrics are key for tracking scientific progress and bridging the current chasm between artificial intelligence research and its translation into practice. However, increasing evidence shows that, particularly in image analysis, metrics are often chosen inadequately. Although taking into account the individual strengths, weaknesses and limitations of validation metrics is a critical prerequisite to making educated choices, the relevant knowledge is currently scattered and poorly accessible to individual researchers. Based on a multistage Delphi process conducted by a multidisciplinary expert consortium as well as extensive community feedback, the present work provides a reliable and comprehensive common point of access to information on pitfalls related to validation metrics in image analysis. Although focused on biomedical image analysis, the addressed pitfalls generalize across application domains and are categorized according to a newly created, domain-agnostic taxonomy. The work serves to enhance global comprehension of a key topic in image analysis validation.


Artificial Intelligence
6.
J Clin Epidemiol ; 165: 111188, 2024 Jan.
Article En | MEDLINE | ID: mdl-37852392

OBJECTIVES: To assess the endorsement of reporting guidelines by high impact factor journals over the period 2017-2022, with a specific focus on the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) statement. STUDY DESIGN AND SETTING: We searched the online 'instructions to authors' of high impact factor medical journals in February 2017 and in January 2022 for any reference to reporting guidelines and TRIPOD in particular. RESULTS: In 2017, 205 out of 337 (61%) journals mentioned any reporting guideline in their instructions to authors and in 2022 this increased to 245 (73%) journals. A reference to TRIPOD was provided by 27 (8%) journals in 2017 and 67 (20%) in 2022. Of those journals mentioning TRIPOD in 2022, 22% provided a link to the TRIPOD website and 60% linked to TRIPOD information on the Enhancing the QUAlity and Transparency Of health Research (EQUATOR) Network website. Twenty-five percent of the journals required adherence to TRIPOD. CONCLUSION: About three-quarters of high-impact medical journals endorse the use of reporting guidelines and 20% endorse TRIPOD. Transparent reporting is important in enhancing the usefulness of health research and endorsement by journals plays a critical role in this.


Periodicals as Topic , Humans , Prognosis , Surveys and Questionnaires
7.
J Clin Epidemiol ; 165: 111206, 2024 Jan.
Article En | MEDLINE | ID: mdl-37925059

OBJECTIVES: Risk of bias assessments are important in meta-analyses of both aggregate and individual participant data (IPD). There is limited evidence on whether and how risk of bias of included studies or datasets in IPD meta-analyses (IPDMAs) is assessed. We review how risk of bias is currently assessed, reported, and incorporated in IPDMAs of test accuracy and clinical prediction model studies and provide recommendations for improvement. STUDY DESIGN AND SETTING: We searched PubMed (January 2018-May 2020) to identify IPDMAs of test accuracy and prediction models, then elicited whether each IPDMA assessed risk of bias of included studies and, if so, how assessments were reported and subsequently incorporated into the IPDMAs. RESULTS: Forty-nine IPDMAs were included. Nineteen of 27 (70%) test accuracy IPDMAs assessed risk of bias, compared to 5 of 22 (23%) prediction model IPDMAs. Seventeen of 19 (89%) test accuracy IPDMAs used Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2), but no tool was used consistently among prediction model IPDMAs. Of IPDMAs assessing risk of bias, 7 (37%) test accuracy IPDMAs and 1 (20%) prediction model IPDMA provided details on the information sources (e.g., the original manuscript, IPD, primary investigators) used to inform judgments, and 4 (21%) test accuracy IPDMAs and 1 (20%) prediction model IPDMA provided information or whether assessments were done before or after obtaining the IPD of the included studies or datasets. Of all included IPDMAs, only seven test accuracy IPDMAs (26%) and one prediction model IPDMA (5%) incorporated risk of bias assessments into their meta-analyses. For future IPDMA projects, we provide guidance on how to adapt tools such as Prediction model Risk Of Bias ASsessment Tool (for prediction models) and QUADAS-2 (for test accuracy) to assess risk of bias of included primary studies and their IPD. CONCLUSION: Risk of bias assessments and their reporting need to be improved in IPDMAs of test accuracy and, especially, prediction model studies. Using recommended tools, both before and after IPD are obtained, will address this.


Data Accuracy , Models, Statistical , Humans , Prognosis , Bias
8.
Diagn Progn Res ; 7(1): 16, 2023 Sep 05.
Article En | MEDLINE | ID: mdl-37667327

BACKGROUND: A previous individual participant data meta-analysis (IPD-MA) of antibiotics for adults with clinically diagnosed acute rhinosinusitis (ARS) showed a marginal overall effect of antibiotics, but was unable to identify patients that are most likely to benefit from antibiotics when applying conventional (i.e. univariable or one-variable-at-a-time) subgroup analysis. We updated the systematic review and investigated whether multivariable prediction of patient-level prognosis and antibiotic treatment effect may lead to more tailored treatment assignment in adults presenting to primary care with ARS. METHODS: An IPD-MA of nine double-blind placebo-controlled trials of antibiotic treatment (n=2539) was conducted, with the probability of being cured at 8-15 days as the primary outcome. A logistic mixed effects model was developed to predict the probability of being cured based on demographic characteristics, signs and symptoms, and antibiotic treatment assignment. Predictive performance was quantified based on internal-external cross-validation in terms of calibration and discrimination performance, overall model fit, and the accuracy of individual predictions. RESULTS: Results indicate that the prognosis with respect to risk of cure could not be reliably predicted (c-statistic 0.58 and Brier score 0.24). Similarly, patient-level treatment effect predictions did not reliably distinguish between those that did and did not benefit from antibiotics (c-for-benefit 0.50). CONCLUSIONS: In conclusion, multivariable prediction based on patient demographics and common signs and symptoms did not reliably predict the patient-level probability of cure and antibiotic effect in this IPD-MA. Therefore, these characteristics cannot be expected to reliably distinguish those that do and do not benefit from antibiotics in adults presenting to primary care with ARS.

9.
BMJ Open ; 13(8): e074984, 2023 08 23.
Article En | MEDLINE | ID: mdl-37612114

INTRODUCTION: The management of type 1 diabetes (T1DM) has undergone significant advancements with the availability of novel technologies, notably continuous and flash glucose monitoring (CGM and FGM, respectively) and hybrid closed loop (HCL) therapy. The dual hormone fully closed loop (DHFCL) approach with insulin and glucagon infusion has shown promising effects in small studies on glycaemic regulation and quality of life in T1DM. METHODS AND ANALYSIS: The Dual Hormone Fully Closed Loop for Type 1 Diabetes (DARE) study is a non-commercial 12-month open-label, two-arm randomised parallel-group trial. The primary aim of this study is to determine the long-term effects on glycaemic control, patient-reported outcome measurements and cost-effectiveness of the DHFCL compared with usual care, that is, HCL or treatment with multiple daily insulin injections+FGM/CGM. We will include 240 adult patients with T1DM in 14 hospitals in the Netherlands. Individuals will be randomised 1:1 to the DHFCL or continuation of their current care. ETHICS AND DISSEMINATION: Ethical approval has been obtained from the Medical Research Ethics Committee NedMec, Utrecht, the Netherlands. Findings will be disseminated through peer-reviewed publications and presentations at local, national and international conferences. TRIAL REGISTRATION NUMBER: NCT05669547.


Diabetes Mellitus, Type 1 , Adult , Humans , Diabetes Mellitus, Type 1/drug therapy , Blood Glucose Self-Monitoring , Netherlands , Quality of Life , Blood Glucose , Insulin/therapeutic use , Randomized Controlled Trials as Topic
10.
Eur Heart J ; 44(32): 3073-3081, 2023 08 22.
Article En | MEDLINE | ID: mdl-37452732

AIMS: Risk stratification is used for decisions regarding need for imaging in patients with clinically suspected acute pulmonary embolism (PE). The aim was to develop a clinical prediction model that provides an individualized, accurate probability estimate for the presence of acute PE in patients with suspected disease based on readily available clinical items and D-dimer concentrations. METHODS AND RESULTS: An individual patient data meta-analysis was performed based on sixteen cross-sectional or prospective studies with data from 28 305 adult patients with clinically suspected PE from various clinical settings, including primary care, emergency care, hospitalized and nursing home patients. A multilevel logistic regression model was built and validated including ten a priori defined objective candidate predictors to predict objectively confirmed PE at baseline or venous thromboembolism (VTE) during follow-up of 30 to 90 days. Multiple imputation was used for missing data. Backward elimination was performed with a P-value <0.10. Discrimination (c-statistic with 95% confidence intervals [CI] and prediction intervals [PI]) and calibration (outcome:expected [O:E] ratio and calibration plot) were evaluated based on internal-external cross-validation. The accuracy of the model was subsequently compared with algorithms based on the Wells score and D-dimer testing. The final model included age (in years), sex, previous VTE, recent surgery or immobilization, haemoptysis, cancer, clinical signs of deep vein thrombosis, inpatient status, D-dimer (in µg/L), and an interaction term between age and D-dimer. The pooled c-statistic was 0.87 (95% CI, 0.85-0.89; 95% PI, 0.77-0.93) and overall calibration was very good (pooled O:E ratio, 0.99; 95% CI, 0.87-1.14; 95% PI, 0.55-1.79). The model slightly overestimated VTE probability in the lower range of estimated probabilities. Discrimination of the current model in the validation data sets was better than that of the Wells score combined with a D-dimer threshold based on age (c-statistic 0.73; 95% CI, 0.70-0.75) or structured clinical pretest probability (c-statistic 0.79; 95% CI, 0.76-0.81). CONCLUSION: The present model provides an absolute, individualized probability of PE presence in a broad population of patients with suspected PE, with very good discrimination and calibration. Its clinical utility needs to be evaluated in a prospective management or impact study. REGISTRATION: PROSPERO ID 89366.


Pulmonary Embolism , Venous Thromboembolism , Adult , Humans , Venous Thromboembolism/diagnosis , Venous Thromboembolism/epidemiology , Prospective Studies , Cross-Sectional Studies , Models, Statistical , Prognosis , Pulmonary Embolism/diagnosis , Pulmonary Embolism/epidemiology , Fibrin Fibrinogen Degradation Products/analysis
11.
J Thromb Haemost ; 21(10): 2873-2883, 2023 10.
Article En | MEDLINE | ID: mdl-37263381

BACKGROUND: In patients clinically suspected of having pulmonary embolism (PE), physicians often rely on intuitive estimation ("gestalt") of PE presence. Although shown to be predictive, gestalt is criticized for its assumed variation across physicians and lack of standardization. OBJECTIVES: To assess the diagnostic accuracy of gestalt in the diagnosis of PE and gain insight into its possible variation. METHODS: We performed an individual patient data meta-analysis including patients suspected of having PE. The primary outcome was diagnostic accuracy of gestalt for the diagnosis of PE, quantified as risk ratio (RR) between gestalt and PE based on 2-stage random-effect log-binomial meta-analysis regression as well as gestalts' sensitivity and specificity. The variability of these measures was explored across different health care settings, publication period, PE prevalence, patient subgroups (sex, heart failure, chronic lung disease, and items of the Wells score other than gestalt), and age. RESULTS: We analyzed 20 770 patients suspected of having PE from 16 original studies. The prevalence of PE in patients with and without a positive gestalt was 28.8% vs 9.1%, respectively. The overall RR was 3.02 (95% CI, 2.35-3.87), and the overall sensitivity and specificity were 74% (95% CI, 68%-79%) and 61% (95% CI, 53%-68%), respectively. Although variation was observed across individual studies (I2, 90.63%), the diagnostic accuracy was consistent across all subgroups and health care settings. CONCLUSION: A positive gestalt was associated with a 3-fold increased risk of PE in suspected patients. Although variation was observed across studies, the RR of gestalt was similar across prespecified subgroups and health care settings, exemplifying its diagnostic value for all patients suspected of having PE.


Physicians , Pulmonary Embolism , Humans , Pulmonary Embolism/diagnosis , Pulmonary Embolism/epidemiology , Sensitivity and Specificity , Male , Female
12.
Stat Med ; 42(19): 3508-3528, 2023 08 30.
Article En | MEDLINE | ID: mdl-37311563

External validation of the discriminative ability of prediction models is of key importance. However, the interpretation of such evaluations is challenging, as the ability to discriminate depends on both the sample characteristics (ie, case-mix) and the generalizability of predictor coefficients, but most discrimination indices do not provide any insight into their respective contributions. To disentangle differences in discriminative ability across external validation samples due to a lack of model generalizability from differences in sample characteristics, we propose propensity-weighted measures of discrimination. These weighted metrics, which are derived from propensity scores for sample membership, are standardized for case-mix differences between the model development and validation samples, allowing for a fair comparison of discriminative ability in terms of model characteristics in a target population of interest. We illustrate our methods with the validation of eight prediction models for deep vein thrombosis in 12 external validation data sets and assess our methods in a simulation study. In the illustrative example, propensity score standardization reduced between-study heterogeneity of discrimination, indicating that between-study variability was partially attributable to case-mix. The simulation study showed that only flexible propensity-score methods (allowing for non-linear effects) produced unbiased estimates of model discrimination in the target population, and only when the positivity assumption was met. Propensity score-based standardization may facilitate the interpretation of (heterogeneity in) discriminative ability of a prediction model as observed across multiple studies, and may guide model updating strategies for a particular target population. Careful propensity score modeling with attention for non-linear relations is recommended.


Benchmarking , Diagnosis-Related Groups , Humans , Computer Simulation
13.
Int J Integr Care ; 23(2): 9, 2023.
Article En | MEDLINE | ID: mdl-37151778

Introduction: Integrated care for patients with atrial fibrillation (AF) in primary care reduced mortality compared to usual care. We assessed the cost-effectiveness of this approach. Methods: Dutch primary care practices were randomised to provide integrated care for AF patients or usual care. A cost-effectiveness analysis was performed from a societal perspective with a 2-year time horizon to estimate incremental costs and Quality Adjusted Life Years (QALYs). A sensitivity analysis was performed, imputing missing questionnaires for a large group of usual care patients. Results: 522 patients from 15 intervention practices were compared to 425 patients from 11 usual care practices. No effect on QALYs was seen, while mean costs indicated a cost reduction between €865 (95% percentile interval (PI) -€5730 to €3641) and €1343 (95% PI -€6534 to €3109) per patient per 2 years. The cost-effectiveness probability ranged between 36% and 54%. In the sensitivity analysis, this increased to 95%-99%. Discussion: Results should be interpreted with caution due to missing information for a large proportion of usual care patients. Conclusion: The higher costs from extra primary care consultations were likely outweighed by cost reductions for other resources, yet this study doesn't give sufficient clarity on the cost-effectiveness of integrated AF care.

15.
J Clin Epidemiol ; 158: 99-110, 2023 06.
Article En | MEDLINE | ID: mdl-37024020

OBJECTIVES: We evaluated the presence and frequency of spin practices and poor reporting standards in studies that developed and/or validated clinical prediction models using supervised machine learning techniques. STUDY DESIGN AND SETTING: We systematically searched PubMed from 01/2018 to 12/2019 to identify diagnostic and prognostic prediction model studies using supervised machine learning. No restrictions were placed on data source, outcome, or clinical specialty. RESULTS: We included 152 studies: 38% reported diagnostic models and 62% prognostic models. When reported, discrimination was described without precision estimates in 53/71 abstracts (74.6% [95% CI 63.4-83.3]) and 53/81 main texts (65.4% [95% CI 54.6-74.9]). Of the 21 abstracts that recommended the model to be used in daily practice, 20 (95.2% [95% CI 77.3-99.8]) lacked any external validation of the developed models. Likewise, 74/133 (55.6% [95% CI 47.2-63.8]) studies made recommendations for clinical use in their main text without any external validation. Reporting guidelines were cited in 13/152 (8.6% [95% CI 5.1-14.1]) studies. CONCLUSION: Spin practices and poor reporting standards are also present in studies on prediction models using machine learning techniques. A tailored framework for the identification of spin will enhance the sound reporting of prediction model studies.


Machine Learning , Humans , Prognosis
16.
Diagn Progn Res ; 7(1): 8, 2023 Apr 04.
Article En | MEDLINE | ID: mdl-37013651

BACKGROUND: The COVID-19 pandemic has a large impact worldwide and is known to particularly affect the older population. This paper outlines the protocol for external validation of prognostic models predicting mortality risk after presentation with COVID-19 in the older population. These prognostic models were originally developed in an adult population and will be validated in an older population (≥ 70 years of age) in three healthcare settings: the hospital setting, the primary care setting, and the nursing home setting. METHODS: Based on a living systematic review of COVID-19 prediction models, we identified eight prognostic models predicting the risk of mortality in adults with a COVID-19 infection (five COVID-19 specific models: GAL-COVID-19 mortality, 4C Mortality Score, NEWS2 + model, Xie model, and Wang clinical model and three pre-existing prognostic scores: APACHE-II, CURB65, SOFA). These eight models will be validated in six different cohorts of the Dutch older population (three hospital cohorts, two primary care cohorts, and a nursing home cohort). All prognostic models will be validated in a hospital setting while the GAL-COVID-19 mortality model will be validated in hospital, primary care, and nursing home settings. The study will include individuals ≥ 70 years of age with a highly suspected or PCR-confirmed COVID-19 infection from March 2020 to December 2020 (and up to December 2021 in a sensitivity analysis). The predictive performance will be evaluated in terms of discrimination, calibration, and decision curves for each of the prognostic models in each cohort individually. For prognostic models with indications of miscalibration, an intercept update will be performed after which predictive performance will be re-evaluated. DISCUSSION: Insight into the performance of existing prognostic models in one of the most vulnerable populations clarifies the extent to which tailoring of COVID-19 prognostic models is needed when models are applied to the older population. Such insight will be important for possible future waves of the COVID-19 pandemic or future pandemics.

17.
J Clin Epidemiol ; 157: 120-133, 2023 05.
Article En | MEDLINE | ID: mdl-36935090

OBJECTIVES: In biomedical research, spin is the overinterpretation of findings, and it is a growing concern. To date, the presence of spin has not been evaluated in prognostic model research in oncology, including studies developing and validating models for individualized risk prediction. STUDY DESIGN AND SETTING: We conducted a systematic review, searching MEDLINE and EMBASE for oncology-related studies that developed and validated a prognostic model using machine learning published between 1st January, 2019, and 5th September, 2019. We used existing spin frameworks and described areas of highly suggestive spin practices. RESULTS: We included 62 publications (including 152 developed models; 37 validated models). Reporting was inconsistent between methods and the results in 27% of studies due to additional analysis and selective reporting. Thirty-two studies (out of 36 applicable studies) reported comparisons between developed models in their discussion and predominantly used discrimination measures to support their claims (78%). Thirty-five studies (56%) used an overly strong or leading word in their title, abstract, results, discussion, or conclusion. CONCLUSION: The potential for spin needs to be considered when reading, interpreting, and using studies that developed and validated prognostic models in oncology. Researchers should carefully report their prognostic model research using words that reflect their actual results and strength of evidence.


Medical Oncology , Research , Humans , Prognosis , Machine Learning
...